Goal Recognition over POMDPs: Inferring the Intention of a POMDP Agent
نویسندگان
چکیده
Plan recognition is the problem of inferring the goals and plans of an agent from partial observations of her behavior. Recently, it has been shown that the problem can be formulated and solved using planners, reducing plan recognition to plan generation. In this work, we extend this model-based approach to plan recognition to the POMDP setting, where actions are stochastic and states are partially observable. The task is to infer a probability distribution over the possible goals of an agent whose behavior results from a POMDP model. The POMDP model is shared between agent and observer except for the true goal of the agent that is hidden to the observer. The observations are action sequencesO that may contain gaps as some or even most of the actions done by the agent may not be observed. We show that the posterior goal distribution P (G|O) can be computed from the value function VG(b) over beliefs b generated by the POMDP planner for each possible goal G. Some extensions of the basic framework are discussed, and a number of experiments are reported.
منابع مشابه
Multiagent Teamwork: Hybrid Approaches
Today within the multiagent community, we see at least four competing methods to building multiagent systems: beliefdesire-intention (BDI), distributed constraint optimization (DCOP), distributed POMDPs, and auctions or game-theoretic methods. While there is exciting progress within each approach, there is a lack of cross-cutting research. This article highlights the various hybrid techniques f...
متن کامل6.845 Final Project: Quantum POMDPs
We present quantum observable Markov decision processes (QOMDPs), the quantum analog of the partially observable Markov decision process (POMDP). In a QOMDP, an agent’s state is represented as a quantum superposition and the agent can choose a superoperator to apply. This is similar to the POMDP belief state, which is a probability distribution over world states and evolves via a stochastic mat...
متن کاملOn Partially Observable MDPs and BDI Models
Decision theoretic planning in ai bymeans of solving Partially ObservableMarkov decision processes (pomdps) has been shown to be both powerful and versatile. However, such approaches are computationally hard and, from a design stance, are not necessarily intuitive for conceptualising many problems. We propose a novel method for solving pomdps, which provides a designer with a more intuitive mea...
متن کاملQuantum POMDPs
We present quantum observable Markov decision processes (QOMDPs), the quantum analogues of partially observable Markov decision processes (POMDPs). In a QOMDP, an agent’s state is represented as a quantum state and the agent can choose a superoperator to apply. This is similar to the POMDP belief state, which is a probability distribution over world states and evolves via a stochastic matrix. W...
متن کاملMulti-Target Detection and Recognition by UAVs Using Online POMDPs
This paper tackles high-level decision-making techniques for robotic missions, which involve both active sensing and symbolic goal reaching, under uncertain probabilistic environments and strong time constraints. Our case study is a POMDP model of an online multi-target detection and recognition mission by an autonomous UAV. The POMDP model of the multitarget detection and recognition problem i...
متن کامل